在查询图像中检索与感兴趣的对象(OOI)在语义上相似的对象具有许多实际用例。一些示例包括修复失败,例如虚假的负面因素/阳性模型或减轻数据集中的类不平衡。有针对性的选择任务需要从大规模的未标记数据池中找到相关数据。在此规模上进行手动开采是不可行的。此外,OOI通常很小,占据图像区域的1%不到1%,被遮挡,并且在混乱的场景中与许多语义上不同的物体共存。现有的语义图像检索方法通常集中在较大尺寸的地理地标的采矿和/或需要额外的标记数据,例如带有相似对象的图像/图像对,用于带有通用对象的挖掘图像。我们在DNN功能空间中提出了一个匹配算法的快速稳固的模板,该模板从一个大的未标记数据池中检索了对象级的语义相似图像。我们将查询图像中OOI周围的区域投射到DNN功能空间以用作模板。这使我们的方法能够专注于OOI的语义,而无需额外的标记数据。在自主驾驶的背景下,我们通过将对象探测器的故障案例作为OOI评估我们的系统进行靶向选择。我们证明了其在具有2.2m图像的大型未标记数据集上的功效,并在采矿中显示出对具有小型OOI的图像的高回忆。我们将我们的方法与众所周知的语义图像检索方法进行比较,该方法也不需要额外的标记数据。最后,我们证明我们的方法是灵活的,并以一种或多种语义上不同的同时发生的OOI无缝地检索图像。
translated by 谷歌翻译
主动学习是一个非常常见但功能强大的框架,用于与人类在循环中的人类迭代和适应性采样子集,目的是实现标签效率。大多数现实世界数据集在类和切片中都有不平衡,并且相应地,数据集的一部分很少见。结果,在设计挖掘这些罕见数据实例的主动学习方法方面已经有很多工作。大多数方法都假设访问包含这些罕见数据实例的一组种子实例。但是,如果发生更极端的稀有性,可以合理地假设这些罕见的数据实例(类或切片)甚至可能在标记的种子集合中存在,并且对主动学习范式的关键需求是有效地发现这些罕见的数据实例。在这项工作中,我们提供了一个主动数据发现框架,该框架可以使用子管的条件增益和下管有条件的相互信息功能有效地挖掘未知的数据切片和类。我们提供了一个一般的算法框架,该框架在许多情况下都起作用,包括图像分类和对象检测,并与未标记集合中存在的稀有类和稀有切片一起使用。与现有的最新活跃学习方法相比,我们的方法表现出显着的准确性和标记效率提高,以积极发现这些稀有类别和切片。
translated by 谷歌翻译
几个射击分类(FSC)需要使用几个(通常为1-5个)数据点的培训模型。事实证明,元学习能够通过培训各种其他分类任务来学习FSC的参数化模型。在这项工作中,我们提出了铂金(使用superodular互信息的半监督模型不可思议的元学习),这是一种新型的半监督模型不合理的元学习框架,使用了子模块化信息(SMI)函数来促进FSC的性能。在元训练期间,使用SMI函数在内部和外循环中利用铂金的数据,并获得元测试的更丰富的元学习参数化。我们在两种情况下研究白金的性能 - 1)未标记的数据点属于与某个插曲的标签集相同的类别集,以及2)在存在不属于的分布类别的地方标记的集合。我们在Miniimagenet,Tieredimagenet和几乎没有Shot-CIFAR100数据集的各种设置上评估了我们的方法。我们的实验表明,铂金优于MAML和半监督的方法,例如用于半监视的FSC的pseduo-Labeling,尤其是对于每个类别的标记示例比例很小。
translated by 谷歌翻译
基于深度神经网络的物体探测器在各种域中取得了巨大的成功,如自主车辆,生物医学成像等。众所周知,他们的成功取决于来自兴趣领域的大量数据。虽然深层模型在整体准确性方面经常表现良好,但它们通常在稀有但关键的数据切片上的性能斗争。例如,像“夜间摩托车”或“夜间摩托车”的数据切片通常很少见但是自动驾驶应用的非常关键的切片,如这种罕见的切片上的假底片可能导致违法的失败和事故。主动学习(AL)是一个着名的范例,可以逐步逐步地和自适应地构建循环中的人类训练数据集。然而,目前基于AL的采集功能并没有充分配备,以解决具有稀有片的真实数据集,因为它们基于图像的不确定性分数或全局描述符。我们提出了Talisman,一种用于使用子模块互信息的稀有切片的目标主动学习或物体检测的新框架。我们的方法使用利用感兴趣区域(ROI)的特征来实用的子模块互信息功能,以有效地靶向并获得具有稀有片的数据点。我们在标准Pascal Voc07 + 12和BDD100K上评估我们的框架,这是一个真实的自动驾驶数据集。我们观察到Talisman在稀有片的平均精度方面优于其他方法,以及地图。
translated by 谷歌翻译
通过选择最具信息丰富的样本,已证明主动学习可用于最小化标记成本。但是,现有的主动学习方法在诸如不平衡或稀有类别的现实方案中不适用于未标记集中的分发数据和冗余。在这项工作中,我们提出了类似的(基于子模块信息措施的主动学习),使用最近提出的子模块信息措施(SIM)作为采集函数的统一主动学习框架。我们认为类似的不仅在标准的主动学习中工作,而且还可以轻松扩展到上面考虑的现实设置,并充当活动学习的一站式解决方案,可以扩展到大型真实世界数据集。凭经验,我们表明,在罕见的课程的情况下,在罕见的阶级和〜5% - 10%的情况下,在罕见的几个图像分类任务的情况下,相似显着优异的活动学习算法像CiFar-10,Mnist和Imagenet。类似于Distil Toolkit的一部分:“https://github.com/decile-team/distil”。
translated by 谷歌翻译
随着数据集大小的不断增加,子集选择技术对于普遍的任务变得越来越重要。通常需要引导子集选择以实现某些探索,其中包括聚焦或针对某些数据点,同时避免他人。这些问题的示例包括:i)目标学习,目标是找到具有罕见类或稀有属性的子集,其中模型表现不佳,II)引导摘要,其中数据(例如,图像集合,文本,文档或视频) )总结了以更快的人类消费与特定的额外用户意图更快。受此类应用程序的动机,我们呈现棱镜,丰富的参数化子模块信息措施。通过小说函数及其参数化,PRISM提供了各种建模能力,该模型能力使得在子集的所需质量之间具有权衡,例如具有一组数据点的分集或表示和相似性/相似性。我们展示了如何应用于上面提到的两个真实问题的棱镜,这需要引导子集选择。在这样做时,我们表明棱镜有趣地概括了一些过去的工作,在其中加强了其广泛的效用。通过对不同数据集的广泛实验,我们展示了棱镜的优越性,在目标学习和引导的图像收集概述中
translated by 谷歌翻译
A popular approach to creating a zero-shot cross-language retrieval model is to substitute a monolingual pretrained language model in the retrieval model with a multilingual pretrained language model such as Multilingual BERT. This multilingual model is fined-tuned to the retrieval task with monolingual data such as English MS MARCO using the same training recipe as the monolingual retrieval model used. However, such transferred models suffer from mismatches in the languages of the input text during training and inference. In this work, we propose transferring monolingual retrieval models using adapters, a parameter-efficient component for a transformer network. By adding adapters pretrained on language tasks for a specific language with task-specific adapters, prior work has shown that the adapter-enhanced models perform better than fine-tuning the entire model when transferring across languages in various NLP tasks. By constructing dense retrieval models with adapters, we show that models trained with monolingual data are more effective than fine-tuning the entire model when transferring to a Cross Language Information Retrieval (CLIR) setting. However, we found that the prior suggestion of replacing the language adapters to match the target language at inference time is suboptimal for dense retrieval models. We provide an in-depth analysis of this discrepancy between other cross-language NLP tasks and CLIR.
translated by 谷歌翻译
Negotiation is one of the crucial abilities in human communication, and there has been a resurgent research interest in negotiation dialogue systems recently, which goal is to empower intelligent agents with such ability that can efficiently help humans resolve conflicts or reach beneficial agreements. Although there have been many explorations in negotiation dialogue systems, a systematic review of this task has to date remained notably absent. To this end, we aim to fill this gap by reviewing contemporary studies in the emerging field of negotiation dialogue systems, covering benchmarks, evaluations, and methodologies. Furthermore, we also discuss potential future directions, including multi-modal, multi-party, and cross-cultural negotiation scenarios. Our goal is to provide the community with a systematic overview of negotiation dialogue systems and to inspire future research.
translated by 谷歌翻译
Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these capabilities have led to widespread adoption, most LLMs are developed by resource-rich organizations and are frequently kept from the public. As a step towards democratizing this powerful technology, we present BLOOM, a 176B-parameter open-access language model designed and built thanks to a collaboration of hundreds of researchers. BLOOM is a decoder-only Transformer language model that was trained on the ROOTS corpus, a dataset comprising hundreds of sources in 46 natural and 13 programming languages (59 in total). We find that BLOOM achieves competitive performance on a wide variety of benchmarks, with stronger results after undergoing multitask prompted finetuning. To facilitate future research and applications using LLMs, we publicly release our models and code under the Responsible AI License.
translated by 谷歌翻译
Real engineering and scientific applications often involve one or more qualitative inputs. Standard Gaussian processes (GPs), however, cannot directly accommodate qualitative inputs. The recently introduced latent variable Gaussian process (LVGP) overcomes this issue by first mapping each qualitative factor to underlying latent variables (LVs), and then uses any standard GP covariance function over these LVs. The LVs are estimated similarly to the other GP hyperparameters through maximum likelihood estimation, and then plugged into the prediction expressions. However, this plug-in approach will not account for uncertainty in estimation of the LVs, which can be significant especially with limited training data. In this work, we develop a fully Bayesian approach for the LVGP model and for visualizing the effects of the qualitative inputs via their LVs. We also develop approximations for scaling up LVGPs and fully Bayesian inference for the LVGP hyperparameters. We conduct numerical studies comparing plug-in inference against fully Bayesian inference over a few engineering models and material design applications. In contrast to previous studies on standard GP modeling that have largely concluded that a fully Bayesian treatment offers limited improvements, our results show that for LVGP modeling it offers significant improvements in prediction accuracy and uncertainty quantification over the plug-in approach.
translated by 谷歌翻译